The original translation in this article is accessing WordNet from Prolog written by Sarah witzig.
Use prolog to operate WordNet WordNet is a semantic network system, and the developer is Princeton University. This article describes the WordNet database of the prolog version and interface predicates in detail.
JWS -- Java WordNet similarity is an open-source project developed by David hope of the University of Sussex to calculate semantic similarity between Java and WordNet.It implements many classical semantic similarityAlgorithm. It is an open-source tool for semantic similarity calculation.
JWS is the Java implementation version of WordNet: similarity (a Perl version of The
natural language processing tool for python, which provides functions to access various WordNet functions. The following lists some common functions:
Get WordNet itself:
From nltk. Corpus import WordNet
Get all the sense of a word, including the various variants of the word:
wordnet.synsets('published')[Synset('print
WordNet Research 2 -- Introduction
Structure
The main relation among words in WordNet is synonymy, as between the words shut and close or car and automobile. synonyms -- words that denote the same concept and are interchangeable in your contexts -- are grouped into unordered sets (synsets ). each of WordNet's 117 000 synsets is linked to other synsets by means of a small number of "Conceptual relatio
dictionary
The dictionary must be said at last. WordNet is a cognitive linguistics-based English dictionary designed by psychologists, linguists and computer engineers at Princeton University. It is not the light that arranges the words in alphabetical order and makes up a "network of words" in terms of the meaning of the words.
Introduction and synonymsMotorcar and automobile are synonyms, which can be studied with the help of WordNet.
(4) wn_ent.plent is the abbreviation of entailment. Only verbs have an implication relationship. For example, snoring implies "sleeping ". The file wn_ent.pl stores all semantic implication relationships, which are predicates: ENT (synset_id_1, synset_id_2 ). The verb set pointed to by synset_id_2 contains a set of synset_id_1. In the working with WordNet section, the predicates find_ent/3 and find_ent_chains/2 are introduced. (5) wn_sim.plsim is the
3. Use WordNet to discuss the basic issues of using WordNet, and give several convenient predicates to improve efficiency. For more documents, see the source program. (1) Create the WordNet database file for indexing Prolog, with a total of 484381 lines of code. Therefore, improving file processing speed is one of the main goals of
WordNet is a library made by Princeton University. You can download it at http://wordnet.princeton.edu. The latest version of Windows is 2.1, which does not seem to have been updated for several years. The installation process is the same as that of General Software. Remember to use the installation directory later. The general installation directory is: C: \ Program Files \ WordNet \ 2.1.
Then under the j
that do not change the meaning of the original verb. For adjectives, the first two Parameters specify one adjective, the last two Parameters specify another adjective, and the latter describes the former. If the synonym set has only one adjective, the word number w_num is 0. For example, the synonym set 300012315 has only one abstract word. Therefore, the predicate SA (300012315,0, 302319830,0 ). SA (300012315,0, 301928363,0 ). represents the invisible words in the collection 302319830 and the
the Wu and Palmer method for semantic related-ness.Let's compare the noun of "ship" and "boat:"= WordNet. Synset(' ship.n.01 ')= WordNet. Synset(' boat.n.01 ')print(W1. Wup_similarity(W2)) 0.9090909090909091= WordNet. Syn
The preparation of the data is the necessary work before the training model, obviously it is also very time-consuming, so in the introductory phase we can use the existing open source image Library to quickly complete the preparatory work: Imagenet
Imagenet is an image database organized according to the WordNet hierarchy (currently only nouns), where each node of the hierarchy is depicted by hundreds of and thousands of images. Currently, each node i
Deep Learning Database Summary
Thanks for the collection.
Source: https://blog.csdn.net/chaipp0607/article/details/71403797
The preparation of the data is necessary to train the model, which is obviously time-consuming, so we can use the existing open source image Library to quickly prepare for the initial work in the introductory phase: ImageNet
Imagenet is an image database organized according to the WordNet hierarchy (currently only nouns), where
The best English dictionary in Linux-general Linux technology-Linux technology and application information. For more information, see the following. StarDict is the best English dictionary for use in Linux.
But how can I add a dictionary to StarDict? I use StarDict for almost 3 ~ I have tried many word libraries in the past four years. The word libraries that have been used and kept up to the present are only the traditional American Dictionary [Dual Solution] and the Oxford dictionary, good
What is jwnl?
Jwnl isAPIFor accessing WordNet-style relational dictionaries. It also provides functionality beyond data access, such as relationship discovery and orphological processing.
Jwnl is an API of WordNet used to access the dictionary in WordNet. jwnl is a Java-based package.
Before using jwnl:
1. Download The wordnetpackage from http://wordnet.princeto
their own images of objects, in the imagenet each category of pictures have their own synset,synset URL: http://www.image-net.org/archive/words.txt, create a local file, save their own needs synset.The following is my Shell batch code:
#!/bin/bash
#save_path = "$"
class_id= "$"
#读取文件内容并保存
id_arr= ($ (cat $class _id | sed '/^#.*\|^$/d '))
id_num=${#id_arr [@]}
n=0 for
var1 in ${id_arr[*]}
do
#$ (wge
', ', ', ' see ', ' You ', ' Dude ', '. '
The word "Mr" is also not separated. NLTK uses the Punktsentencetokenizer of the Punkt module, which is part of the nltk.tokenize. And this tokenizer is trained to work in multiple languages.
Non-English tokenize
You can specify the language when Tokenize:
From nltk.tokenize Import sent_tokenizemytext = "Bonjour m. Adam, comment allez-vous?" J ' espère que tout va bien. Aujourd ' hui est un bon jour. " Print (Sent_tokenize (mytext, "French"))
The outpu
/get_ilsvrc_aux.shThe training data and the validation data entry are described in the text train.txt and Val.txt lists all the files and their labels. Note that we use a label index that is different from the ILSVRC Devkit: We sort the synset names in ASCII order, and then mark them from 0 to 999. You can see the synset/name mapping in Synset_words.txt.You may need to adjust the image to 256x256 in advance
, chat logs, movie reviews, and a whole lot More.Now, we ' re going-talk about accessing these documents via NLTK. As can see, these is mostly text documents, so you could just use normal Python code to open and read documents. That said, the NLTK module have a few nice methods for handling the corpus, so could find it useful to use their metholog Y. Here's an example of us opening the Gutenberg Bible, and reading the first few lines:FromNltk.TokenizeImportSent_tokenize, PunktsentencetokenizerFr
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.